natural environment
Reconstructing Animals and the Wild
Kulits, Peter, Black, Michael J., Zuffi, Silvia
The idea of 3D reconstruction as scene understanding is foundational in computer vision. Reconstructing 3D scenes from 2D visual observations requires strong priors to disambiguate structure. Much work has been focused on the anthropocentric, which, characterized by smooth surfaces, coherent normals, and regular edges, allows for the integration of strong geometric inductive biases. Here, we consider a more challenging problem where such assumptions do not hold: the reconstruction of natural scenes containing trees, bushes, boulders, and animals. While numerous works have attempted to tackle the problem of reconstructing animals in the wild, they have focused solely on the animal, neglecting environmental context. This limits their usefulness for analysis tasks, as animals exist inherently within the 3D world, and information is lost when environmental factors are disregarded. We propose a method to reconstruct natural scenes from single images. We base our approach on recent advances leveraging the strong world priors ingrained in Large Language Models and train an autoregressive model to decode a CLIP embedding into a structured compositional scene representation, encompassing both animals and the wild (RAW). To enable this, we propose a synthetic dataset comprising one million images and thousands of assets. Our approach, having been trained solely on synthetic data, generalizes to the task of reconstructing animals and their environments in real-world images. We will release our dataset and code to encourage future research at https://raw.is.tue.mpg.de/
- Asia > Japan > Honshū > Chūbu > Ishikawa Prefecture > Kanazawa (0.05)
- North America > United States > Massachusetts (0.04)
- Europe > Switzerland (0.04)
- (4 more...)
Matched Filtering based LiDAR Place Recognition for Urban and Natural Environments
Joseph, Therese, Fischer, Tobias, Milford, Michael
Place recognition is an important task within autonomous navigation, involving the re-identification of previously visited locations from an initial traverse. Unlike visual place recognition (VPR), LiDAR place recognition (LPR) is tolerant to changes in lighting, seasons, and textures, leading to high performance on benchmark datasets from structured urban environments. However, there is a growing need for methods that can operate in diverse environments with high performance and minimal training. In this paper, we propose a handcrafted matching strategy that performs roto-translation invariant place recognition and relative pose estimation for both urban and unstructured natural environments. Our approach constructs Birds Eye View (BEV) global descriptors and employs a two-stage search using matched filtering -- a signal processing technique for detecting known signals amidst noise. Extensive testing on the NCLT, Oxford Radar, and WildPlaces datasets consistently demonstrates state-of-the-art (SoTA) performance across place recognition and relative pose estimation metrics, with up to 15% higher recall than previous SoTA.
- Oceania > Australia > Queensland > Brisbane (0.14)
- North America > United States > Michigan (0.04)
Evaluating saliency scores in point clouds of natural environments by learning surface anomalies
Arav, Reuma, Wittich, Dennis, Rottensteiner, Franz
In recent years, three-dimensional point clouds are used increasingly to document natural environments. Each dataset contains a diverse set of objects, at varying shapes and sizes, distributed throughout the data and intricately intertwined with the topography. Therefore, regions of interest are difficult to find and consequent analyses become a challenge. Inspired from visual perception principles, we propose to differentiate objects of interest from the cluttered environment by evaluating how much they stand out from their surroundings, i.e., their geometric salience. Previous saliency detection approaches suggested mostly handcrafted attributes for the task. However, such methods fail when the data are too noisy or have high levels of texture. Here we propose a learning-based mechanism that accommodates noise and textured surfaces. We assume that within the natural environment any change from the prevalent surface would suggest a salient object. Thus, we first learn the underlying surface and then search for anomalies within it. Initially, a deep neural network is trained to reconstruct the surface. Regions where the reconstructed part deviates significantly from the original point cloud yield a substantial reconstruction error, signifying an anomaly, i.e., saliency. We demonstrate the effectiveness of the proposed approach by searching for salient features in various natural scenarios, which were acquired by different acquisition platforms. We show the strong correlation between the reconstruction error and salient objects.
- Europe > Austria > Vienna (0.04)
- Asia > Middle East > Israel (0.04)
- Europe > Switzerland (0.04)
- (4 more...)
Galibr: Targetless LiDAR-Camera Extrinsic Calibration Method via Ground Plane Initialization
Song, Wonho, Oh, Minho, Lee, Jaeyoung, Myung, Hyun
With the rapid development of autonomous driving and SLAM technology, the performance of autonomous systems using multimodal sensors highly relies on accurate extrinsic calibration. Addressing the need for a convenient, maintenance-friendly calibration process in any natural environment, this paper introduces Galibr, a fully automatic targetless LiDAR-camera extrinsic calibration tool designed for ground vehicle platforms in any natural setting. The method utilizes the ground planes and edge information from both LiDAR and camera inputs, streamlining the calibration process. It encompasses two main steps: an initial pose estimation algorithm based on ground planes (GP-init), and a refinement phase through edge extraction and matching. Our approach significantly enhances calibration performance, primarily attributed to our novel initial pose estimation method, as demonstrated in unstructured natural environments, including on the KITTI dataset and the KAIST quadruped dataset.
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Robots > Autonomous Vehicles (0.49)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
ChatGPTest: opportunities and cautionary tales of utilizing AI for questionnaire pretesting
Olivos, Francisco, Liu, Minhui
Pretesting involves a small-scale trial of data collection procedures, aiming to assess them. It is a standard practice in both academic and applied research (Grimm 2010), and the output of the pretest is usually the feedback offered by interviewers on how to improve procedures and questions. The rapid advancements in generative artificial intelligence (GAI) have opened up new avenues for enhancing various aspects of research, including the design and evaluation of survey questionnaires. AI technologies like large language models (LLMs) have demonstrated remarkable potential in generating human-like text, offering a promising approach to pretesting survey instruments. This article explores the use of GPT models as a tool for pretesting survey questionnaires. Illustrated with two applications, it suggests incorporating GPT feedback as an additional stage before human pretesting, potentially reducing successive iterations. However, the article emphasizes the indispensable role of researchers' judgment in implementing AIgenerated feedback. GPT is an LLM that utilizes advanced algorithms to generate texts that mimic the syntax, semantics, and grammar of human writing, which are approximated by statistical patterns learned from training data (for a technical review, see OpenAI 2023). Like most of the LLMs, GPT models predict the next word in a sequence based on the preceding words.
- Asia > China > Hong Kong (0.05)
- North America > United States > New Jersey (0.04)
- North America > United States > Kentucky > Butler County (0.04)
- Research Report (1.00)
- Questionnaire & Opinion Survey (1.00)
Towards Long-term Robotics in the Wild
Hausler, Stephen, Griffiths, Ethan, Ramezani, Milad, Moghadam, Peyman
In this paper, we emphasise the critical importance of large-scale datasets for advancing field robotics capabilities, particularly in natural environments. While numerous datasets exist for urban and suburban settings, those tailored to natural environments are scarce. Our recent benchmarks WildPlaces and WildScenes address this gap by providing synchronised image, lidar, semantic and accurate 6-DoF pose information in forest-type environments. We highlight the multi-modal nature of this dataset and discuss and demonstrate its utility in various downstream tasks, such as place recognition and 2D and 3D semantic segmentation tasks.
- Oceania > Australia > Queensland > Brisbane (0.05)
- Europe > Germany > Baden-Württemberg > Freiburg (0.04)
Linear Decision Rule as Aspiration for Simple Decision Heuristics
Several attempts to understand the success of simple decision heuristics have examined heuristics as an approximation to a linear decision rule. This research has identified three environmental structures that aid heuristics: dominance, cumulative dominance, and noncompensatoriness. This paper develops these ideas further and examines their empirical relevance in 51 natural environments. The results show that all three structures are prevalent, making it possible for simple rules to reach, and occasionally exceed, the accuracy of the linear decision rule, using less information and less computation.
- North America > United States > New York (0.05)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.05)
- Europe > Germany > Berlin (0.04)
Watch a plant-inspired robot grow towards light like a vine
A robot that can grow around trees or rocks like a vine could be used to make buildings or measure pollution in hard-to-reach natural environments. Vine-like robots aren't new, but they are often designed to rely on just a single sense to grow upwards, such as heat or light, which means they don't work as well in some settings as others. Emanuela Del Dottore at the Italian Institute of Technology and her colleagues have developed a new version, called FiloBot, that can use light, shade or gravity as a guide. It grows by coiling a plastic filament into a cylindrical shape, adding new layers to its body just behind the head that contains the sensors. "Our robot has an embedded microcontroller that can process multiple stimuli and direct the growth at a precise location, the tip, ensuring the body structure is preserved," she says.
WildScenes: A Benchmark for 2D and 3D Semantic Segmentation in Large-scale Natural Environments
Vidanapathirana, Kavisha, Knights, Joshua, Hausler, Stephen, Cox, Mark, Ramezani, Milad, Jooste, Jason, Griffiths, Ethan, Mohamed, Shaheer, Sridharan, Sridha, Fookes, Clinton, Moghadam, Peyman
Recent progress in semantic scene understanding has primarily been enabled by the availability of semantically annotated bi-modal (camera and lidar) datasets in urban environments. However, such annotated datasets are also needed for natural, unstructured environments to enable semantic perception for applications, including conservation, search and rescue, environment monitoring, and agricultural automation. Therefore, we introduce WildScenes, a bi-modal benchmark dataset consisting of multiple large-scale traversals in natural environments, including semantic annotations in high-resolution 2D images and dense 3D lidar point clouds, and accurate 6-DoF pose information. The data is (1) trajectory-centric with accurate localization and globally aligned point clouds, (2) calibrated and synchronized to support bi-modal inference, and (3) containing different natural environments over 6 months to support research on domain adaptation. Our 3D semantic labels are obtained via an efficient automated process that transfers the human-annotated 2D labels from multiple views into 3D point clouds, thus circumventing the need for expensive and time-consuming human annotation in 3D. We introduce benchmarks on 2D and 3D semantic segmentation and evaluate a variety of recent deep-learning techniques to demonstrate the challenges in semantic segmentation in natural environments. We propose train-val-test splits for standard benchmarks as well as domain adaptation benchmarks and utilize an automated split generation technique to ensure the balance of class label distributions. The data, evaluation scripts and pretrained models will be released upon acceptance at https://csiro-robotics.github.io/WildScenes.
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Robots (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.86)
BotanicGarden: A high-quality and large-scale robot navigation dataset in challenging natural environments
Liu, Yuanzhi, Fu, Yujia, Qin, Minghui, Xu, Yufeng, Xu, Baoxin, Chen, Fengdong, Goossens, Bart, Yu, Hongwei, Liu, Chun, Chen, Long, Tao, Wei, Zhao, Hui
The rapid developments of mobile robotics and autonomous navigation over the years are largely empowered by public datasets for testing and upgrading, such as SLAM and localization tasks. Impressive demos and benchmark results have arisen, indicating the establishment of a mature technical framework. However, from the view point of real-world deployments, there are still critical defects of robustness in challenging environments, especially in large-scale, GNSS-denied, textural-monotonous, and unstructured scenarios. To meet the pressing validation demands in such scope, we build a novel challenging robot navigation dataset in a large botanic garden of more than 48000m2. Comprehensive sensors are employed, including high-res/rate stereo Gray&RGB cameras, rotational and forward 3D LiDARs, and low-cost and industrial-grade IMUs, all of which are well calibrated and accurately hardware-synchronized. An all-terrain wheeled robot is configured to mount the sensor suite and provide odometry data. A total of 32 long and short sequences of 2.3 million images are collected, covering scenes of thick woods, riversides, narrow paths, bridges, and grasslands that rarely appeared in previous resources. Excitedly, both highly-accurate ego-motions and 3D map ground truth are provided, along with fine-annotated vision semantics. Our goal is to contribute a high-quality dataset to advance robot navigation and sensor fusion research to a higher level.
- Asia > China > Shanghai > Shanghai (0.04)
- Asia > China > Heilongjiang Province > Harbin (0.04)
- North America > United States > Texas > Kleberg County (0.04)
- (9 more...)